Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher.
Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?
Some links on this page may take you to non-federal websites. Their policies may differ from this site.
-
Free, publicly-accessible full text available June 1, 2026
-
Free, publicly-accessible full text available March 1, 2026
-
In this paper, we study efficient approximate sampling for probability distributions known up to normalization constants. We specifically focus on a problem class arising in Bayesian inference for large-scale inverse problems in science and engineering applications. The computational challenges we address with the proposed methodology are: (i) the need for repeated evaluations of expensive forward models; (ii) the potential existence of multiple modes; and (iii) the fact that gradient of, or adjoint solver for, the forward model might not be feasible. While existing Bayesian inference methods meet some of these challenges individually, we propose a framework that tackles all three systematically. Our approach builds upon the Fisher–Rao gradient flow in probability space, yielding a dynamical system for probability densities that converges towards the target distribution at a uniform exponential rate. This rapid convergence is advantageous for the computational burden outlined in (i). We apply Gaussian mixture approximations with operator splitting techniques to simulate the flow numerically; the resulting approximation can capture multiple modes thus addressing (ii). Furthermore, we employ the Kalman methodology to facilitate a derivative-free update of these Gaussian components and their respective weights, addressing the issue in (iii). The proposed methodology results in an efficient derivative-free posterior approximation method, flexible enough to handle multi-modal distributions: Gaussian Mixture Kalman Inversion (GMKI). The effectiveness of GMKI is demonstrated both theoretically and numerically in several experiments with multimodal target distributions, including proof-of-concept and two-dimensional examples, as well as a large-scale application: recovering the Navier–Stokes initial condition from solution data at positive times.more » « less
-
Abstract We study uniformly random lozenge tilings of general simply connected polygons. Under a technical assumption that is presumably generic with respect to polygon shapes, we show that the local statistics around a cusp point of the arctic curve converge to the Pearcey process. This verifies the widely predicted universality of edge statistics in the cusp case. Together with the smooth and tangent cases proved by Aggarwal‐Huang and Aggarwal‐Gorin, these are believed to be the three types of edge statistics that can arise in a generic polygon. Our proof is via a local coupling of the random tiling with nonintersecting Bernoulli random walks (NBRW). To leverage this coupling, we establish an optimal concentration estimate for the tiling height function around the cusp. As another step and also a result of potential independent interest, we show that the local statistics of NBRW around a cusp converge to the Pearcey process when the initial configuration consists of two parts with proper density growth, via careful asymptotic analysis of the determinantal formulas.more » « less
-
Abstract Consider the normalized adjacency matrices of randomd‐regular graphs onNvertices with fixed degree . We prove that, with probability for any , the following two properties hold as provided that : (i) The eigenvalues are close to the classical eigenvalue locations given by the Kesten–McKay distribution. In particular, the extremal eigenvalues are concentrated with polynomial error bound inN, that is, . (ii) All eigenvectors of randomd‐regular graphs are completely delocalized.more » « less
-
Abstract We consider Bayesian inference for large-scale inverse problems, where computational challenges arise from the need for repeated evaluations of an expensive forward model. This renders most Markov chain Monte Carlo approaches infeasible, since they typically require O ( 1 0 4 ) model runs, or more. Moreover, the forward model is often given as a black box or is impractical to differentiate. Therefore derivative-free algorithms are highly desirable. We propose a framework, which is built on Kalman methodology, to efficiently perform Bayesian inference in such inverse problems. The basic method is based on an approximation of the filtering distribution of a novel mean-field dynamical system, into which the inverse problem is embedded as an observation operator. Theoretical properties are established for linear inverse problems, demonstrating that the desired Bayesian posterior is given by the steady state of the law of the filtering distribution of the mean-field dynamical system, and proving exponential convergence to it. This suggests that, for nonlinear problems which are close to Gaussian, sequentially computing this law provides the basis for efficient iterative methods to approximate the Bayesian posterior. Ensemble methods are applied to obtain interacting particle system approximations of the filtering distribution of the mean-field model; and practical strategies to further reduce the computational and memory cost of the methodology are presented, including low-rank approximation and a bi-fidelity approach. The effectiveness of the framework is demonstrated in several numerical experiments, including proof-of-concept linear/nonlinear examples and two large-scale applications: learning of permeability parameters in subsurface flow; and learning subgrid-scale parameters in a global climate model. Moreover, the stochastic ensemble Kalman filter and various ensemble square-root Kalman filters are all employed and are compared numerically. The results demonstrate that the proposed method, based on exponential convergence to the filtering distribution of a mean-field dynamical system, is competitive with pre-existing Kalman-based methods for inverse problems.more » « less
An official website of the United States government

Full Text Available